46 research outputs found

    LO-Net: Deep Real-time Lidar Odometry

    Full text link
    We present a novel deep convolutional network pipeline, LO-Net, for real-time lidar odometry estimation. Unlike most existing lidar odometry (LO) estimations that go through individually designed feature selection, feature matching, and pose estimation pipeline, LO-Net can be trained in an end-to-end manner. With a new mask-weighted geometric constraint loss, LO-Net can effectively learn feature representation for LO estimation, and can implicitly exploit the sequential dependencies and dynamics in the data. We also design a scan-to-map module, which uses the geometric and semantic information learned in LO-Net, to improve the estimation accuracy. Experiments on benchmark datasets demonstrate that LO-Net outperforms existing learning based approaches and has similar accuracy with the state-of-the-art geometry-based approach, LOAM

    Point2Node: Correlation Learning of Dynamic-Node for Point Cloud Feature Modeling

    Full text link
    Fully exploring correlation among points in point clouds is essential for their feature modeling. This paper presents a novel end-to-end graph model, named Point2Node, to represent a given point cloud. Point2Node can dynamically explore correlation among all graph nodes from different levels, and adaptively aggregate the learned features. Specifically, first, to fully explore the spatial correlation among points for enhanced feature description, in a high-dimensional node graph, we dynamically integrate the node's correlation with self, local, and non-local nodes. Second, to more effectively integrate learned features, we design a data-aware gate mechanism to self-adaptively aggregate features at the channel level. Extensive experiments on various point cloud benchmarks demonstrate that our method outperforms the state-of-the-art.Comment: AAAI2020(oral

    HSC4D: Human-centered 4D Scene Capture in Large-scale Indoor-outdoor Space Using Wearable IMUs and LiDAR

    Full text link
    We propose Human-centered 4D Scene Capture (HSC4D) to accurately and efficiently create a dynamic digital world, containing large-scale indoor-outdoor scenes, diverse human motions, and rich interactions between humans and environments. Using only body-mounted IMUs and LiDAR, HSC4D is space-free without any external devices' constraints and map-free without pre-built maps. Considering that IMUs can capture human poses but always drift for long-period use, while LiDAR is stable for global localization but rough for local positions and orientations, HSC4D makes both sensors complement each other by a joint optimization and achieves promising results for long-term capture. Relationships between humans and environments are also explored to make their interaction more realistic. To facilitate many down-stream tasks, like AR, VR, robots, autonomous driving, etc., we propose a dataset containing three large scenes (1k-5k m2m^2) with accurate dynamic human motions and locations. Diverse scenarios (climbing gym, multi-story building, slope, etc.) and challenging human activities (exercising, walking up/down stairs, climbing, etc.) demonstrate the effectiveness and the generalization ability of HSC4D. The dataset and code are available at http://www.lidarhumanmotion.net/hsc4d/.Comment: 10 pages, 8 figures, CVPR202

    RF-Net: An End-to-End Image Matching Network based on Receptive Field

    Full text link
    This paper proposes a new end-to-end trainable matching network based on receptive field, RF-Net, to compute sparse correspondence between images. Building end-to-end trainable matching framework is desirable and challenging. The very recent approach, LF-Net, successfully embeds the entire feature extraction pipeline into a jointly trainable pipeline, and produces the state-of-the-art matching results. This paper introduces two modifications to the structure of LF-Net. First, we propose to construct receptive feature maps, which lead to more effective keypoint detection. Second, we introduce a general loss function term, neighbor mask, to facilitate training patch selection. This results in improved stability in descriptor training. We trained RF-Net on the open dataset HPatches, and compared it with other methods on multiple benchmark datasets. Experiments show that RF-Net outperforms existing state-of-the-art methods.Comment: 9 pages, 6 figure

    Semantic Labeling of Mobile LiDAR Point Clouds via Active Learning and Higher Order MRF

    Get PDF
    【Abstract】Using mobile Light Detection and Ranging point clouds to accomplish road scene labeling tasks shows promise for a variety of applications. Most existing methods for semantic labeling of point clouds require a huge number of fully supervised point cloud scenes, where each point needs to be manually annotated with a specific category. Manually annotating each point in point cloud scenes is labor intensive and hinders practical usage of those methods. To alleviate such a huge burden of manual annotation, in this paper, we introduce an active learning method that avoids annotating the whole point cloud scenes by iteratively annotating a small portion of unlabeled supervoxels and creating a minimal manually annotated training set. In order to avoid the biased sampling existing in traditional active learning methods, a neighbor-consistency prior is exploited to select the potentially misclassified samples into the training set to improve the accuracy of the statistical model. Furthermore, lots of methods only consider short-range contextual information to conduct semantic labeling tasks, but ignore the long-range contexts among local variables. In this paper, we use a higher order Markov random field model to take into account more contexts for refining the labeling results, despite of lacking fully supervised scenes. Evaluations on three data sets show that our proposed framework achieves a high accuracy in labeling point clouds although only a small portion of labels is provided. Moreover, comparative experiments demonstrate that our proposed framework is superior to traditional sampling methods and exhibits comparable performance to those fully supervised models.10.13039/501100001809-National Natural Science Foundation of China; Collaborative Innovation Center of Haixi Government Affairs Big Data Sharin

    Pairwise registration of TLS point clouds by deep multi-scale local features

    Get PDF
    Abstract(#br)Because of the mechanism of TLS system, noise, outliers, various occlusions, varying cloud densities, etc. inevitably exist in the collection of TLS point clouds. To achieve automatic TLS point cloud registration, many methods, based on the hand-crafted features of keypoints, have been proposed. Despite significant progress, the current methods still face great challenges in accomplishing TLS point cloud registration. In this paper, we propose a multi-scale neural network to learn local shape descriptors for establishing correspondences between pairwise TLS point clouds. To train our model, data augmentation, developed on pairwise semi-synthetic 3D local patches, is to extend our network to be robust to rotation transformation. Then, based on varying local neighborhoods, multi-scale subnetworks are constructed and fused to learn robust local features. Experimental results demonstrate that our proposed method successfully registers two TLS point clouds and outperforms state-of-the-art methods. Besides, our learned descriptors are invariant to translation and tolerant to changes in rotation

    Image-based orchard insect automated identification and classification method

    No full text
    Department of Entomology, Michigan State University; USDA-ARS Post-harvest Lab, East Lansing, MIInsect identification and classification is time-consuming work requiring expert knowledge for integrated pest management in orchards. An image-based automated insect identification and classification method is described in the paper. The complete method includes three models. An invariant local feature model was built for insect identification and classification using affine invariant local features; a global feature model was built for insect identification and classification using 54 global features; and a hierarchical combination model was proposed based on local feature and global feature models to combine advantages of the two models and increase performance. The three models were applied and tested for insect classification on eight insect species from pest colonies and orchards. The hierarchical combination model yielded better performance over global and local models. Moreover, to study the pose change of insects on traps and the hypothesis that an optimal time to acquire and image after landing exists, advanced analysis on time-dependent pose change of insects on traps is included in this study. The experimental results on field insect image classification with field-based images for training achieved the classification rate of 86.6% when testing with the combination model. This demonstrates the image-based insect identification and classification method could be a potential way for automated insect classification in integrated pest management. (C) 2012 Elsevier B.V. All rights reserved

    Pose estimation-dependent identification method for field moth images using deep learning architecture

    No full text
    Due to the varieties of moth poses and cluttered background, traditional methods for automated identification of on-trap moths suffer problems of incomplete feature extraction and misidentification. A novel pose estimation-dependent automated identification method using deep learning architecture is proposed in this paper for on-trap field moth sample images. To deal with cluttered background and uneven illumination, two-level automated moth segmentation was created for separating moth sample images from each trap image. Moth pose was then estimated in terms of either top view or side view. Suitable combinations of texture, colour, shape and local features were extracted for further moth description. Finally, the improved pyramidal stacked de-noising auto-encoder (IpSDAE) architecture was proposed to build a deep neural network for moth identification. The experimental results on 762 field moth samples by 10-fold cross-validation achieved a good identification accuracy of 96.9%, and indicated that the deployment of the proposed pose estimation process is effective for automated moth identification
    corecore